An Incremental Gradient(-Projection) Method with Momentum Term and Adaptive Stepsize Rule

نویسنده

  • Paul Tseng
چکیده

We consider an incremental gradient method with momentum term for minimizing the sum of continuously differentiable functions. This method uses a new adaptive stepsize rule that decreases the stepsize whenever sufficient progress is not made. We show that if the gradients of the functions are bounded and Lipschitz continuous over a certain level set, then every cluster point of the iterates generated by the method is a stationary point. In addition, if the gradient of the functions have a certain growth property, then the method is either linearly convergent in some sense or the stepsizes are bounded away from zero. The new stepsize rule is much in the spirit of heuristic learning rules used in practice for training neural networks via backpropagation. As such, the new stepsize rule may suggest improvements on existing learning rules. Finally, extension of the method and the convergence results to constrained minimization is discussed, as are some implementation issues and numerical experience.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Note on the Gradient Projection Method with Exact Stepsize Rule *1)

In this paper, we give some convergence results on the gradient projection method with exact stepsize rule for solving the minimization problem with convex constraints. Especially, we show that if the objective function is convex and its gradient is Lipschitz continuous, then the whole sequence of iterations produced by this method with bounded exact stepsizes converges to a solution of the con...

متن کامل

A globally convergent incremental Newton method

Motivated by machine learning problems over large data sets and distributed optimization over networks, we develop and analyze a new method called incremental Newton method for minimizing the sum of a large number of strongly convex functions. We show that our method is globally convergent for a variable stepsize rule. We further show that under a gradient growth condition, convergence rate is ...

متن کامل

A self-adaptive gradient projection algorithm for the nonadditive traffic equilibrium problem

Gradient projection (GP) algorithm has been shown as an efficient algorithm for solving the traditional traffic equilibrium problem with additive route costs. Recently, GP has been extended to solve the nonadditive traffic equilibrium problem (NaTEP), in which the cost incurred on each route is not just a simple sum of the link costs on that route. However, choosing an appropriate stepsize, whi...

متن کامل

Convergence of a Grassmannian Gradient Descent Algorithm for Subspace Estimation From Undersampled Data

Subspace learning and matrix factorization problems have a great many applications in science and engineering, and efficient algorithms are critical as dataset sizes continue to grow. Many relevant problem formulations are non-convex, and in a variety of contexts it has been observed that solving the non-convex problem directly is not only efficient but reliably accurate. We discuss convergence...

متن کامل

An adaptive pressure correction method without spurious velocities for diffuse-interface models of incompressible flows

In this article, we propose to study two issues associated with the use of the incremental projection method for solving the incompressible Navier-Stokes equation. The first one is the combination of this time splitting algorithm with an adaptive local refinement method. The second one is the reduction of spurious velocities due to the right-hand side of the momentum balance. We propose a new v...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • SIAM Journal on Optimization

دوره 8  شماره 

صفحات  -

تاریخ انتشار 1998